Skip to content

[ROCm] Updated compilation_pipeline_rocm.cc#40377

Closed
zoranjovanovic-ns wants to merge 1 commit intoopenxla:mainfrom
ROCm:rocm-update-triton-pipeline
Closed

[ROCm] Updated compilation_pipeline_rocm.cc#40377
zoranjovanovic-ns wants to merge 1 commit intoopenxla:mainfrom
ROCm:rocm-update-triton-pipeline

Conversation

@zoranjovanovic-ns
Copy link
Copy Markdown
Contributor

📝 Summary of Changes
Updated compilation_pipeline_rocm.cc to be aligned with corresponding compiletion.py

🎯 Justification
Using the same layout of Triton passes as triton compiler

🚀 Kind of Contribution
Please remove what does not apply: ♻️ Cleanup

📊 Benchmark (for Performance Improvements)
Please measure and include speedups for one of the public HLOs in
compiler/xla/tools/benchmarks/hlo/.

🧪 Unit Tests:
triton/support_test

🧪 Execution Tests:
What execution tests were added? For example, a new optimization should be
tested with an end-to-end execution test triggering the optimization and
asserting correctness. Please provide test cases running with at most 2 GPUs.

@i-chaochen i-chaochen added the AMD-GPU XLA on AMD GPU label Apr 3, 2026
@github-project-automation github-project-automation bot moved this to Pending Google Review in OpenXLA:GPU - AMD/ROCm Apr 3, 2026
copybara-service bot pushed a commit that referenced this pull request Apr 8, 2026
Imported from GitHub PR #40377

📝 Summary of Changes
Updated compilation_pipeline_rocm.cc to be aligned with corresponding compiletion.py

🎯 Justification
Using the same layout of Triton passes as triton compiler

🚀 Kind of Contribution
Please remove what does not apply: ♻️ Cleanup

📊 Benchmark (for Performance Improvements)
Please measure and include speedups for one of the public HLOs in
`compiler/xla/tools/benchmarks/hlo/`.

🧪 Unit Tests:
triton/support_test

🧪 Execution Tests:
What execution tests were added? For example, a new optimization should be
tested with an end-to-end execution test triggering the optimization and
asserting correctness. Please provide test cases running with at most 2 GPUs.

Copybara import of the project:

--
ab40190 by Zoran Jovanovic <zjovanov@amd.com>:

[ROCm] Updated compilation_pipeline_rocm.cc

Merging this change closes #40377

FUTURE_COPYBARA_INTEGRATE_REVIEW=#40377 from ROCm:rocm-update-triton-pipeline ab40190
PiperOrigin-RevId: 896382030
copybara-service bot pushed a commit to tensorflow/tensorflow that referenced this pull request Apr 8, 2026
Imported from GitHub PR openxla/xla#40377

📝 Summary of Changes
Updated compilation_pipeline_rocm.cc to be aligned with corresponding compiletion.py

🎯 Justification
Using the same layout of Triton passes as triton compiler

🚀 Kind of Contribution
Please remove what does not apply: ♻️ Cleanup

📊 Benchmark (for Performance Improvements)
Please measure and include speedups for one of the public HLOs in
`compiler/xla/tools/benchmarks/hlo/`.

🧪 Unit Tests:
triton/support_test

🧪 Execution Tests:
What execution tests were added? For example, a new optimization should be
tested with an end-to-end execution test triggering the optimization and
asserting correctness. Please provide test cases running with at most 2 GPUs.

Copybara import of the project:

--
ab4019047932affe889edf67651c6b0ced228f95 by Zoran Jovanovic <zjovanov@amd.com>:

[ROCm] Updated compilation_pipeline_rocm.cc

Merging this change closes #40377

FUTURE_COPYBARA_INTEGRATE_REVIEW=openxla/xla#40377 from ROCm:rocm-update-triton-pipeline ab4019047932affe889edf67651c6b0ced228f95
PiperOrigin-RevId: 896382030
copybara-service bot pushed a commit that referenced this pull request Apr 8, 2026
Imported from GitHub PR #40377

📝 Summary of Changes
Updated compilation_pipeline_rocm.cc to be aligned with corresponding compiletion.py

🎯 Justification
Using the same layout of Triton passes as triton compiler

🚀 Kind of Contribution
Please remove what does not apply: ♻️ Cleanup

📊 Benchmark (for Performance Improvements)
Please measure and include speedups for one of the public HLOs in
`compiler/xla/tools/benchmarks/hlo/`.

🧪 Unit Tests:
triton/support_test

🧪 Execution Tests:
What execution tests were added? For example, a new optimization should be
tested with an end-to-end execution test triggering the optimization and
asserting correctness. Please provide test cases running with at most 2 GPUs.

Copybara import of the project:

--
ab40190 by Zoran Jovanovic <zjovanov@amd.com>:

[ROCm] Updated compilation_pipeline_rocm.cc

Merging this change closes #40377

FUTURE_COPYBARA_INTEGRATE_REVIEW=#40377 from ROCm:rocm-update-triton-pipeline ab40190
PiperOrigin-RevId: 896382030
copybara-service bot pushed a commit that referenced this pull request Apr 8, 2026
Imported from GitHub PR #40377

📝 Summary of Changes
Updated compilation_pipeline_rocm.cc to be aligned with corresponding compiletion.py

🎯 Justification
Using the same layout of Triton passes as triton compiler

🚀 Kind of Contribution
Please remove what does not apply: ♻️ Cleanup

📊 Benchmark (for Performance Improvements)
Please measure and include speedups for one of the public HLOs in
`compiler/xla/tools/benchmarks/hlo/`.

🧪 Unit Tests:
triton/support_test

🧪 Execution Tests:
What execution tests were added? For example, a new optimization should be
tested with an end-to-end execution test triggering the optimization and
asserting correctness. Please provide test cases running with at most 2 GPUs.

Copybara import of the project:

--
ab40190 by Zoran Jovanovic <zjovanov@amd.com>:

[ROCm] Updated compilation_pipeline_rocm.cc

Merging this change closes #40377

FUTURE_COPYBARA_INTEGRATE_REVIEW=#40377 from ROCm:rocm-update-triton-pipeline ab40190
PiperOrigin-RevId: 896382030
@copybara-service copybara-service bot closed this in 3ffb98e Apr 8, 2026
@github-project-automation github-project-automation bot moved this from Pending Google Review to Done in OpenXLA:GPU - AMD/ROCm Apr 8, 2026
copybara-service bot pushed a commit to tensorflow/tensorflow that referenced this pull request Apr 8, 2026
Imported from GitHub PR openxla/xla#40377

📝 Summary of Changes
Updated compilation_pipeline_rocm.cc to be aligned with corresponding compiletion.py

🎯 Justification
Using the same layout of Triton passes as triton compiler

🚀 Kind of Contribution
Please remove what does not apply: ♻️ Cleanup

📊 Benchmark (for Performance Improvements)
Please measure and include speedups for one of the public HLOs in
`compiler/xla/tools/benchmarks/hlo/`.

🧪 Unit Tests:
triton/support_test

🧪 Execution Tests:
What execution tests were added? For example, a new optimization should be
tested with an end-to-end execution test triggering the optimization and
asserting correctness. Please provide test cases running with at most 2 GPUs.

Copybara import of the project:

--
ab4019047932affe889edf67651c6b0ced228f95 by Zoran Jovanovic <zjovanov@amd.com>:

[ROCm] Updated compilation_pipeline_rocm.cc

Merging this change closes #40377

PiperOrigin-RevId: 896406714
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

AMD-GPU XLA on AMD GPU

Projects

Archived in project

Development

Successfully merging this pull request may close these issues.

3 participants